Gdańsk
- North America > Canada > Ontario > Toronto (0.14)
- Europe > Poland > Masovia Province > Warsaw (0.04)
- Europe > Poland > Pomerania Province > Gdańsk (0.04)
- Education (0.46)
- Information Technology (0.46)
The toddler who survived a 54-degree body temperature
Humans aren't built for the cold, but have survived frigid temperatures in some amazing cases. Breakthroughs, discoveries, and DIY tips sent six days a week. Winter is not for the faint of heart. In New York City, skyscrapers turn Manhattan into a series of freezing wind tunnels. In Sapporo, Japan, the snowfall is almost 200 inches each winter. Even so, humans have developed plenty of clever ways to wait out the cold. But what would happen if instead of bundling up inside with a hot chocolate, you were left in the frigid cold--just how cold can humans get and recover?
- North America > United States > New York (0.25)
- Asia > Japan > Hokkaidō > Hokkaidō Prefecture > Sapporo (0.24)
- Europe > Russia > Central Federal District > Moscow Oblast > Moscow (0.04)
- (3 more...)
- Health & Medicine > Therapeutic Area (0.73)
- Health & Medicine > Diagnostic Medicine > Vital Signs (0.45)
- Information Technology > Communications > Mobile (0.42)
- Information Technology > Artificial Intelligence (0.35)
Model-Based Ranking of Source Languages for Zero-Shot Cross-Lingual Transfer
Ebrahimi, Abteen, Wiemerslage, Adam, von der Wense, Katharina
We present NN-Rank, an algorithm for ranking source languages for cross-lingual transfer, which leverages hidden representations from multilingual models and unlabeled target-language data. We experiment with two pretrained multilingual models and two tasks: part-of-speech tagging (POS) and named entity recognition (NER). We consider 51 source languages and evaluate on 56 and 72 target languages for POS and NER, respectively. When using in-domain data, NN-Rank beats state-of-the-art baselines that leverage lexical and linguistic features, with average improvements of up to 35.56 NDCG for POS and 18.14 NDCG for NER. As prior approaches can fall back to language-level features if target language data is not available, we show that NN-Rank remains competitive using only the Bible, an out-of-domain corpus available for a large number of languages. Ablations on the amount of unlabeled target data show that, for subsets consisting of as few as 25 examples, NN-Rank produces high-quality rankings which achieve 92.8% of the NDCG achieved using all available target data for ranking.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Europe > Hungary > Csongrád-Csanád County > Szeged (0.04)
- Europe > Poland > Pomerania Province > Gdańsk (0.04)
- (15 more...)
Lecture Notes on Verifying Graph Neural Networks
In these lecture notes, we first recall the connection between graph neural networks, Weisfeiler-Lehman tests and logics such as first-order logic and graded modal logic. We then present a modal logic in which counting modalities appear in linear inequalities in order to solve verification tasks on graph neural networks. We describe an algorithm for the satisfiability problem of that logic. It is inspired from the tableau method of vanilla modal logic, extended with reasoning in quantifier-free fragment Boolean algebra with Presburger arithmetic.
Integrating Domain Knowledge into Process Discovery Using Large Language Models
Norouzifar, Ali, Kourani, Humam, Dees, Marcus, van der Aalst, Wil
Process discovery aims to derive process models from event logs, providing insights into operational behavior and forming a foundation for conformance checking and process improvement. However, models derived solely from event data may not accurately reflect the real process, as event logs are often incomplete or affected by noise, and domain knowledge, an important complementary resource, is typically disregarded. As a result, the discovered models may lack reliability for downstream tasks. We propose an interactive framework that incorporates domain knowledge, expressed in natural language, into the process discovery pipeline using Large Language Models (LLMs). Our approach leverages LLMs to extract declarative rules from textual descriptions provided by domain experts. These rules are used to guide the IMr discovery algorithm, which recursively constructs process models by combining insights from both the event log and the extracted rules, helping to avoid problematic process structures that contradict domain knowledge. The framework coordinates interactions among the LLM, domain experts, and a set of backend services. We present a fully implemented tool that supports this workflow and conduct an extensive evaluation of multiple LLMs and prompt engineering strategies. Our empirical study includes a case study based on a real-life event log with the involvement of domain experts, who assessed the usability and effectiveness of the framework.
- Europe > Austria > Vienna (0.14)
- Europe > Portugal (0.04)
- North America > United States > New Mexico > Santa Fe County > Santa Fe (0.04)
- (15 more...)
- Workflow (1.00)
- Research Report (1.00)
- Overview (1.00)
When Models Lie, We Learn: Multilingual Span-Level Hallucination Detection with PsiloQA
Rykov, Elisei, Petrushina, Kseniia, Savkin, Maksim, Olisov, Valerii, Vazhentsev, Artem, Titova, Kseniia, Panchenko, Alexander, Konovalov, Vasily, Belikova, Julia
Hallucination detection remains a fundamental challenge for the safe and reliable deployment of large language models (LLMs), especially in applications requiring factual accuracy. Existing hallucination benchmarks often operate at the sequence level and are limited to English, lacking the fine-grained, multilingual supervision needed for a comprehensive evaluation. In this work, we introduce PsiloQA, a large-scale, multilingual dataset annotated with span-level hallucinations across 14 languages. PsiloQA is constructed through an automated three-stage pipeline: generating question-answer pairs from Wikipedia using GPT-4o, eliciting potentially hallucinated answers from diverse LLMs in a no-context setting, and automatically annotating hallucinated spans using GPT-4o by comparing against golden answers and retrieved context. We evaluate a wide range of hallucination detection methods -- including uncertainty quantification, LLM-based tagging, and fine-tuned encoder models -- and show that encoder-based models achieve the strongest performance across languages. Furthermore, PsiloQA demonstrates effective cross-lingual generalization and supports robust knowledge transfer to other benchmarks, all while being significantly more cost-efficient than human-annotated datasets. Our dataset and results advance the development of scalable, fine-grained hallucination detection in multilingual settings.
Ultra-Fast Language Generation via Discrete Diffusion Divergence Instruct
Zheng, Haoyang, Liu, Xinyang, Kong, Cindy Xiangrui, Jiang, Nan, Hu, Zheyuan, Luo, Weijian, Deng, Wei, Lin, Guang
Fast and high-quality language generation is the holy grail that people pursue in the age of AI. In this work, we introduce Discrete Diffusion Divergence Instruct (DiDi-Instruct), a training-based method that initializes from a pre-trained (masked) discrete diffusion language model (dLLM) and distills a few-step student for fast generation. The resulting DiDi-Instruct model achieves comparable or superior performance to its dLLM teacher and the GPT-2 baseline while enabling up to 64$\times$ acceleration. The theoretical foundation of DiDi-Instruct is a novel framework based on integral KL-divergence minimization, which yields a practical training algorithm. We further introduce grouped reward normalization, intermediate-state matching, and the reward-guided ancestral sampler that significantly improve training stability, model coverage, and inference quality. On OpenWebText, DiDi-Instruct achieves perplexity from 62.2 (8 NFEs) to 18.4 (128 NFEs), which outperforms prior accelerated dLLMs and GPT-2 baseline. These gains come with a negligible entropy loss (around $1\%$) and reduce additional training wall-clock time by more than $20\times$ compared to competing dLLM distillation methods. We further validate the robustness and effectiveness of DiDi-Instruct through extensive ablation studies, model scaling, and the generation of discrete protein sequences. In conclusion, DiDi-Instruct is an efficient yet effective distillation method, enabling language generation in the blink of an eye. We will release both code and models at github.com/haoyangzheng-ai/didi-instruct.
- Asia > Russia (0.93)
- Asia > India (0.14)
- North America > United States > Virginia (0.04)
- (33 more...)
- Research Report > New Finding (1.00)
- Personal > Interview (0.92)
Analysing Python Machine Learning Notebooks with Moose
Mignard, Marius, Costiou, Steven, Anquetil, Nicolas, Etien, Anne
Machine Learning (ML) code, particularly within notebooks, often exhibits lower quality compared to traditional software. Bad practices arise at three distinct levels: general Python coding conventions, the organizational structure of the notebook itself, and ML-specific aspects such as reproducibility and correct API usage. However, existing analysis tools typically focus on only one of these levels and struggle to capture ML-specific semantics, limiting their ability to detect issues. This paper introduces Vespucci Linter, a static analysis tool with multi-level capabilities, built on Moose and designed to address this challenge. Leveraging a metamodeling approach that unifies the notebook's structural elements with Python code entities, our linter enables a more contextualized analysis to identify issues across all three levels. We implemented 22 linting rules derived from the literature and applied our tool to a corpus of 5,000 notebooks from the Kaggle platform. The results reveal violations at all levels, validating the relevance of our multi-level approach and demonstrating Vespucci Linter's potential to improve the quality and reliability of ML development in notebook environments.
- North America > United States > New York > New York County > New York City (0.05)
- Europe > Poland > Pomerania Province > Gdańsk (0.04)
- Europe > France > Hauts-de-France > Nord > Lille (0.04)
Exploring the Impact of Generative Artificial Intelligence on Software Development in the IT Sector: Preliminary Findings on Productivity, Efficiency and Job Security
Bonin, Anton Ludwig, Smolinski, Pawel Robert, Winiarski, Jacek
This study investigates the impact of Generative AI on software development within the IT sector through a mixed-method approach, utilizing a survey developed based on expert interviews. The preliminary results of an ongoing survey offer early insights into how Generative AI reshapes personal productivity, organizational efficiency, adoption, business strategy and job insecurity. The findings reveal that 97% of IT workers use Generative AI tools, mainly ChatGPT. Participants report significant personal productivity gain and perceive organizational efficiency improvements that correlate positively with Generative AI adoption by their organizations (r = .470, p < .05). However, increased organizational adoption of AI strongly correlates with heightened employee job security concerns (r = .549, p < .001). Key adoption challenges include inaccurate outputs (64.2%), regulatory compliance issues (58.2%) and ethical concerns (52.2%). This research offers early empirical insights into Generative AI's economic and organizational implications.
- Europe > Poland > Pomerania Province > Gdańsk (0.05)
- North America > United States > District of Columbia > Washington (0.04)
- North America > United States > New York (0.04)
- (4 more...)
- Questionnaire & Opinion Survey (1.00)
- Overview (0.93)
- Research Report > New Finding (0.69)
- Research Report > Experimental Study (0.46)
Effective Clustering for Large Multi-Relational Graphs
Lin, Xiaoyang, Jiang, Runhao, Yang, Renchi
Multi-relational graphs (MRGs) are an expressive data structure for modeling diverse interactions/relations among real objects (i.e., nodes), which pervade extensive applications and scenarios. Given an MRG G with N nodes, partitioning the node set therein into K disjoint clusters (MRGC) is a fundamental task in analyzing MRGs, which has garnered considerable attention. However, the majority of existing solutions towards MRGC either yield severely compromised result quality by ineffective fusion of heterogeneous graph structures and attributes, or struggle to cope with sizable MRGs with millions of nodes and billions of edges due to the adoption of sophisticated and costly deep learning models. In this paper, we present DEMM and DEMM+, two effective MRGC approaches to address the limitations above. Specifically, our algorithms are built on novel two-stage optimization objectives, where the former seeks to derive high-caliber node feature vectors by optimizing the multi-relational Dirichlet energy specialized for MRGs, while the latter minimizes the Dirichlet energy of clustering results over the node affinity graph. In particular, DEMM+ achieves significantly higher scalability and efficiency over our based method DEMM through a suite of well-thought-out optimizations. Key technical contributions include (i) a highly efficient approximation solver for constructing node feature vectors, and (ii) a theoretically-grounded problem transformation with carefully-crafted techniques that enable linear-time clustering without explicitly materializing the NxN dense affinity matrix. Further, we extend DEMM+ to handle attribute-less MRGs through non-trivial adaptations. Extensive experiments, comparing DEMM+ against 20 baselines over 11 real MRGs, exhibit that DEMM+ is consistently superior in terms of clustering quality measured against ground-truth labels, while often being remarkably faster.
- North America > United States > District of Columbia > Washington (0.05)
- Asia > China > Hong Kong (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- (5 more...)